$ curl -X POST "https://scrape.infatica.io/" -d '{"api_key": "API_KEY", "url": "URL"}' <!DOCTYPE html> <html lang="en"> <DOCTYPE html> <meta charset=”utf-8”> ... $ Click here to see full documentation...
Superior Web Scraper API: Handling Headless (JS Rendering) and Rotation Proxy
Web Scraping Should Be Simple
When designing Scraper API, we had a simple goal: Make it efficient for power users – and intuitive for home users. Data extraction sounds really complicated: Will this website break? What if JavaScript rendering or geotargeting is required? What about anti-scraping systems like reCAPTCHA and Cloudflare? Scraper API solves all of these problems – and finally makes web scraping simple.
- Built for scalability: our large pool of residential proxies suits any large-scale project.
- Free 24/7 support: our specialists are ready to troubleshoot any of your technical problems.
- Easy to use: we handle the technical side of the scraping workflow (e.g. proxy management) and save your time.
- Reliable and stable: we’ve designed this product with performance and connection stability in mind.
curl -X POST "https://scrape.infatica.io/" -d '{"api_key": "API_KEY", "url": "https://www.google.com"}'
const axios = require('axios')
const options = {
method: 'POST',
responseType: 'json',
data: {
url: 'TARGET_URL',
api_key: 'API_KEY',
country_code: 'gb' // country_code parameter is optional
},
url: 'https://scrape.infatica.io'
}
axios(options)
.then((result) => {
console.log(result)
})
.catch((err) => {
console.error(err)
})
import requests
import json
req = requests.post('https://scrape.infatica.io/', data = json.dumps({
'url': 'TARGET_URL',
'api_key': 'API_KEY',
'country_code': 'us', # country_code parameter is optional
'headers': { # headers parameter is optional
'user_header_1': 'header1_value',
'user_header_2': 'header2_value'
}
}))
content = json.loads(req.content)
print(content)
Extract Data From Dynamic Websites
Dynamic content is the backbone of modern tech platforms: real-time price changes, product updates, messaging, efficient pagination, and much more. The constant flow of this data is enabled by web browsers’ JavaScript rendering capabilities – but this code can be problematic for parsers to process correctly.
Infatica’s Scraper API addresses this issue thanks to its robust rendering engine which features full JavaScript rendering, Ajax support, and pagination handlers – the latter allows us to parse both single-page and multi-page websites and their components. Working together, these features enable you to scrape all URLs on any popular website without a single data point missing.
Export Your Data In CSV, XLSX, and JSON
Scraping data is half the job done – now you need a proper method of processing it. File formats are used to organize data in a machine-readable way, allowing human users to view and edit them easily. Popular response formats include CSV and XLSX for arranging tabular data (e.g. as Excel spreadsheets), and JSON for organizing data in web applications.
Infatica’s Scraper API supports all of these response formats, providing you full control over your data organization workflow. Export data in XLSX spreadsheets and CSV files to analyze in Excel or use JSON for easy API and webhook access.
Get Structured Data Fast – Without Roadblocks
Thousands of companies are investing their resources into web data extraction – and data owners respond by adopting anti-scraping systems like reCAPTCHA and Cloudflare. These security measures are designed to distinguish between real users and web scrapers, which they attempt to via a set of factors. One of these factors is the IP address: If it raises suspicion, using a web scraper becomes much harder due to regular IP bans.
Scraper API solves this problem using Infatica’s own residential proxy network, which makes the requests your crawlers send appear human-like – and this helps to avoid triggering CAPTCHAs and IP address bans. Thanks to Infatica’s proxies and reliable server infrastructure, Scraper API achieves a high request success rate, low response time, maximum uptime, and best performance.
High request success rate
Low response time
Maximum uptime
Best performance
Features of our advanced data collection suite
Millions of proxies & IPs
Infatica Scraper utilizes our own network of residential IP addresses across dozens of global ISPs, supporting real devices, smart retries, and proxy rotation.
100+ global locations
Choose from 100+ global locations via powerful geotargeting to send your web scraping API requests from – or simply use random geo-targets from a set of major cities all across the globe.
Robust infrastructure
Scrape the web at scale at an unparalleled speed and enjoy advanced features like concurrent API requests, CAPTCHA solving, browser support and JS rendering.
Free and premium options
Are you here to test the API without any commitments? Sign up for our Free Plan. If you ever need more advanced access, premium pricing plans start at $19.99 per month.
What Customers Say about us
“Excellent service and customer care! For the first time, I bought a proxy on Infatica last year. I immediately bribed the responsiveness of the support, they helped me choose the tariff I needed for my purposes, and they actively answered all the questions later. There are no complaints about the proxies themselves; everything suits me)”
Use Scraper by yourself
Get Free TrialSmall Project
Access to premium proxies and up to 250k monthly requests.
- JS Rendering - Yes
- Json parsing - Yes
- Built-in residential proxy - Yes
- US & EU Geotargeting - Yes
- Threads - 10
- Ticket support - Yes
Medium Project
Access to premium proxies and up to 1 million monthly requests.
- JS Rendering - Yes
- Json parsing - Yes
- Built-in residential proxy - Yes
- US & EU Geotargeting - Yes
- Threads - 50
- Ticket support - Yes
Large Project
Business level - everything we have and up to 3 million monthly requests.
- JS Rendering - Yes
- Json parsing - Yes
- Built-in residential proxy - Yes
- US & EU Geotargeting - Yes
- Threads - 100
- Ticket support - Yes
Enterprise
Enterpise level - everything we have and up to custom monthly requests.
- JS Rendering - Yes
- Json parsing - Yes
- Built-in residential proxy - Yes
- US & EU Geotargeting - Yes
- Threads - Customize
- Ticket support - Yes
Commitment-Free Trial
Trial request count: 5,000 requests. Trial duration: 7 days.
Frequently Asked Questions
-
Generally, yes: As of 2022, intellectual property laws are not explicitly prohibiting web scraping. A recent decision of the US Supreme Court states: If a website provides publicly available data and doesn’t require authorization, accessing this data is legal.
-
Web scraping means automated collection of website data. The keyword here is automated: Although you can save web data manually, specialized software (e.g. scrapers and crawlers) enable this process to scale across thousands of websites – and this software can run efficiently even on a regular home computer.
-
Upon collecting data, you can analyze it to explain trends and make educated guesses. Some good examples of these products include price aggregation platforms, e-commerce businesses, search engine optimization services, fraud protection software, and more.
-
The simplest method is using software with point and click interfaces: You click at the given website’s element (e.g. a table) and the program saves its data. Power users create more advanced scrapers that use the browser to read the website’s code, providing more control over their web scraping workflow.
-
In data collection, Python is arguably the fan favorite thanks to its wide range of pre-made libraries for networking and file operations. Still, other languages (e.g. JavaScript) have web scraping utilities of their own, so choosing something other than Python shouldn’t present any problems.
-
Home versions of Google Chrome or Microsoft’s Edge browser aren’t suitable for scraping, so their specialized versions are used instead. They are called headless browsers because they lack the graphical interface that we normally use to browse websites. Some popular examples of these browsers include Headless Chrome, Headless Firefox, and PhantomJS.
-
In general, intellectual property laws do not consider scraping platforms like Google or Amazon to be illegal. Google’s Terms of Service, however, prohibits automated access; the consequences of breaking ToS may include IP blocks (making web scraper’s job harder), but Google hasn’t actually sued any company for scraping its data.
-
Yes, but legal action isn’t the likely outcome in most web scraping scenarios. For this to happen, you have to extract data from a website and republish it. Conversely, transforming this data in a meaningful way (e.g. to create a price aggregator) falls under the fair use doctrine and is OK.
-
This largely depends on the scale of operation – a simple scraper for a project will cost less. If you don’t want to run the web scraper from your home computer, virtual machines are available for rent starting at just a few dollars per month. Additionally, you will require proxies to protect the data miner’s requests: Similarly, their pricing starts at $3-4 per GB.
-
Yes, but there are some caveats. Amazon’s data is public, so accessing and collecting using data miners is legal. To keep a pipeline that involves Amazon’s data legal, you need to transform it so that it offers a new perspective – a good example is price monitoring website.
-
Proxy rotation is a feature of Infatica’s proxy network: It monitors its entire pool of IP addresses and detects if the given address has been blocked by the target website. If this happens, the blocked IP is replaced with a new one, keeping the scraping pipeline uninterrupted and making web data extraction quicker.
-
It is a powerful scraper that allows you to crawl various websites at a large scale, in real time. As a tool for professional data collection, Scraper API makes web scraping easier via automating different processes like bulk scrape jobs, scheduled scrape processes, own customized extraction rule, and more.
-
There is barely the “best” user agent – you only need one that isn’t deemed suspicious by the target server. Most common user agents for web scraping include combinations of Chrome 101.0 + Windows 10 (9.9% of users), Firefox 100.0 + Windows 10 (8.1% of users), and Chrome 101.0 + macOS (5.1% of users.)
-
Technically, yes – as long as the data is actually public and isn’t locked behind an authorization gateway. Thanks to Scraper API’s JavaScript rendering capabilities, you can extract data from any popular website – search engines, ecommerce platforms, knowledge bases, forums, newspapers, file archives, social media platforms, aggregators, and more.
-
In most cases, yes. However, some APIs feature free and premium options, with the latter typically lifting the platform’s scraping restrictions. More importantly, large-scale scraping projects come with associated costs: renting a virtual machine and purchasing proxies are a must – without these upgrades, the scraping pipeline may be inefficient.
-
Infatica Scraper offers a set of response formats for exporting and organizing scraped data: JSON and HTML. You can use them to arrange data in a tabular manner or feed it to your web application via an API.